international cooperation
Alignment and Safety in Large Language Models: Safety Mechanisms, Training Paradigms, and Emerging Challenges
Lu, Haoran, Fang, Luyang, Zhang, Ruidong, Li, Xinliang, Cai, Jiazhang, Cheng, Huimin, Tang, Lin, Liu, Ziyu, Sun, Zeliang, Wang, Tao, Zhang, Yingchuan, Zidan, Arif Hassan, Xu, Jinwen, Yu, Jincheng, Yu, Meizhi, Jiang, Hanqi, Gong, Xilin, Luo, Weidi, Sun, Bolun, Chen, Yongkai, Ma, Terry, Wu, Shushan, Zhou, Yifan, Chen, Junhao, Xiang, Haotian, Zhang, Jing, Jahin, Afrar, Ruan, Wei, Deng, Ke, Pan, Yi, Wang, Peilong, Li, Jiahui, Liu, Zhengliang, Zhang, Lu, Zhao, Lin, Liu, Wei, Zhu, Dajiang, Xing, Xin, Dou, Fei, Zhang, Wei, Huang, Chao, Liu, Rongjie, Zhang, Mengrui, Liu, Yiwen, Sun, Xiaoxiao, Lu, Qin, Xiang, Zhen, Zhong, Wenxuan, Liu, Tianming, Ma, Ping
Due to the remarkable capabilities and growing impact of large language models (LLMs), they have been deeply integrated into many aspects of society. Thus, ensuring their alignment with human values and intentions has emerged as a critical challenge. This survey provides a comprehensive overview of practical alignment techniques, training protocols, and empirical findings in LLM alignment. We analyze the development of alignment methods across diverse paradigms, characterizing the fundamental trade-offs between core alignment objectives. Our analysis shows that while supervised fine-tuning enables basic instruction-following, preference-based methods offer more flexibility for aligning with nuanced human intent. We discuss state-of-the-art techniques, including Direct Preference Optimization (DPO), Constitutional AI, brain-inspired methods, and alignment uncertainty quantification (AUQ), highlighting their approaches to balancing quality and efficiency. We review existing evaluation frameworks and benchmarking datasets, emphasizing limitations such as reward misspecification, distributional robustness, and scalable oversight. We summarize strategies adopted by leading AI labs to illustrate the current state of practice. We conclude by outlining open problems in oversight, value pluralism, robustness, and continuous alignment. This survey aims to inform both researchers and practitioners navigating the evolving landscape of LLM alignment.
- North America > United States > Georgia > Clarke County > Athens (0.14)
- North America > United States > Arizona > Pima County > Tucson (0.14)
- Europe > Austria > Vienna (0.14)
- (28 more...)
- Research Report > Promising Solution (1.00)
- Research Report > New Finding (1.00)
- Overview (1.00)
- Information Technology > Security & Privacy (1.00)
- Health & Medicine > Therapeutic Area > Neurology (1.00)
- Government > Military (1.00)
- (7 more...)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Uncertainty > Bayesian Inference (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)
- (2 more...)
TechScape: What we learned from the global AI summit in South Korea
What does success look like for the second global AI summit? As the great and good of the industry (and me) gathered last week at the Korea Institute of Science and Technology, a sprawling hilltop campus in eastern Seoul, that was the question I kept asking myself. If we're ranking the event by the quantity of announcements generated, then it's a roaring success. In less than 24 hours – starting with a virtual "leader's summit" at 8pm and ending with a joint press conference with the South Korean and British science and technology ministers – I counted no fewer than six agreements, pacts, pledges and statements, all demonstrating the success of the event in getting people around the table to hammer out a deal. The first 16 companies have signed up to voluntary artificial intelligence safety standards introduced at the Bletchley Park summit, Rishi Sunak has said on the eve of the follow-up event in Seoul.
- Asia > South Korea > Seoul > Seoul (0.55)
- Europe > United Kingdom > England > Buckinghamshire > Milton Keynes (0.25)
- North America > United States (0.05)
- (2 more...)
France to host next AI safety summit as European nations jockey for tech leadership
AI expert Marva Bailer tells Fox News Digital how the open availability of artificial intelligence can have negative impacts and talks potential federal legislation to control it. European nations continue to jockey for leadership on artificial intelligence (AI), with Paris announcing it will host the next safety summit shortly after Britain hosted the first one. "The first edition of the Artificial Intelligence Security Summit, organized by the United Kingdom, provides an opportunity to develop international cooperation in the field of security, a crucial issue for the years to come. It was, therefore, natural for France to host the second edition of this summit," French Minister Delegate for the Digital Economy Jean-Noël Barrot said in a press release. The future of AI remains up for grabs, with many nations trying to position themselves at the forefront of the race.
- Europe > France (0.77)
- Europe > United Kingdom > England > Buckinghamshire > Milton Keynes (0.10)
- Asia > China (0.09)
- (2 more...)
- Information Technology > Security & Privacy (0.55)
- Government > Regional Government > Europe Government (0.52)
- Media > News (0.39)
Why China's Involvement in the U.K. AI Safety Summit Was So Significant
As delegates from around the world and leaders from the tech industry gathered in the U.K. for the first ever AI Safety Summit, there appeared to be harmony between officials of historical rivals China and the U.S. On Wednesday, U.S. Secretary of Commerce Gina Raimondo and Chinese Vice Minister of Science and Technology Wu Zhaohui shared a stage at the opening plenary of the U.K. AI Safety Summit. Later that day, the U.S. and China were two of 29 countries to sign the Bletchley Declaration, which affirmed the risks AI poses and commits to international cooperation to identify and mitigate those risks, a crucial initial step to establishing regulation in the future. But simmering beneath these shows of cooperation is an undercurrent of increasing tension between the two AI superpowers. Tech supremacy has been a hallmark of the tensions between the U.S. and China in recent years. In 2017, on the heels of an impressive breakthrough in artificial intelligence by Google Deepmind, China made AI progress a priority with its New Generation AI Development Plan.
- North America > United States (1.00)
- Europe > Russia (0.05)
- Asia > Russia (0.05)
- (4 more...)
- Government > Regional Government > North America Government > United States Government (1.00)
- Government > Commerce (0.90)
The US, not China, should take the lead on AI
Senior fellow at the Gatestone Institute Gordon Chang joined'Cavuto Live' to discuss the U.S.'s relationship with China amid the highly anticipated G20 Summit. Emerging technologies like artificial intelligence (AI) should be used as "tools of opportunity, not as weapons of oppression," President Biden remarked recently. But this exhortation makes his subsequent vow to work directly with "our competitors" to harness the power of AI "for good" all the more curious. Working with our competitors, like China, would only empower the Chinese Communist Party (CCP) to write the rules of the road for AI. And we don't want China in the driver's seat.
- North America > United States (1.00)
- Asia > Russia (0.15)
- South America > Ecuador (0.05)
- (10 more...)
- Government > Military (1.00)
- Government > Regional Government > North America Government > United States Government (0.91)
- Government > Regional Government > Asia Government > China Government (0.69)
How the U.N. Plans to Shape the Future of AI
As the United Nations General Assembly gathered this week in New York, the U.N. Secretary-General's envoy on technology, Amandeep Gill, hosted an event titled Governing AI for Humanity, where participants discussed the risks that AI might pose and the challenges of achieving international cooperation on artificial intelligence. Secretary-General António Guterres and Gill have said they believe that a new U.N. agency will be required to help the world cooperate in managing this powerful technology. But the issues that the new entity would seek to address and its structure are yet to be determined, and some observers say that ambitious plans for global cooperation like this rarely get the required support of powerful nations. Gill has led efforts to make advanced forms of technology safer before. He was chair of the Group of Governmental Experts of the Convention on Certain Conventional Weapons when the Campaign to Stop Killer Robots, which sought to compel governments to outlaw the development of lethal autonomous weapons systems, failed to gain traction with global superpowers including the U.S. and Russia.
- Europe > Russia (0.35)
- Asia > Russia (0.35)
- North America > United States > New York (0.25)
- (3 more...)
- Government > Military (0.68)
- Energy > Power Industry > Utilities > Nuclear (0.50)
We Can Prevent AI Disaster Like We Prevented Nuclear Catastrophe
On 16th July 1945 the world changed forever. The Manhattan Project's'Trinity' test, directed by Robert Oppenheimer, endowed humanity for the first time with the ability to wipe itself out: an atomic bomb had been successfully detonated 210 miles south of Los Alamos, New Mexico. On 6th August 1945 the bomb was dropped on Hiroshima and three days later, Nagasaki-- unleashing unprecedented destructive power. The end of World War II brought a fragile peace, overshadowed by this new, existential threat. While nuclear technology promised an era of abundant energy, it also launched us into a future where nuclear war could lead to the end of our civilization.
- North America > United States > New Mexico > Los Alamos County > Los Alamos (0.25)
- Asia > Japan > Kyūshū & Okinawa > Kyūshū > Nagasaki Prefecture > Nagasaki (0.25)
- Asia > Japan > Honshū > Chūgoku > Hiroshima Prefecture > Hiroshima (0.25)
- Europe > United Kingdom (0.05)
Explore the possibility of advancing climate negotiations on the basis of regional trade organizations: A study based on RICE-N
Climate issues have become more and more important now. Although global governments have made some progress, we are still facing the truth that the prospect of international cooperation is not clear at present. Due to the limitations of the Integrated assessment models (IAMs) model, it is difficult to simulate the dynamic negotiation process. Therefore, using deep learning to build a new agents based model (ABM) might can provide new theoretical support for climate negotiations. Building on the RICE-N model, this work proposed an approach to climate negotiations based on existing trade groups. Simulation results show that the scheme has a good prospect.
US launches artificial intelligence military use initiative - ABC News
The United States launched an initiative Thursday promoting international cooperation on the responsible use of artificial intelligence and autonomous weapons by militaries, seeking to impose order on an emerging technology that has the potential to change the way war is waged. "As a rapidly changing technology, we have an obligation to create strong norms of responsible behavior concerning military uses of AI and in a way that keeps in mind that applications of AI by militaries will undoubtedly change in the coming years," Bonnie Jenkins, the State Department's under secretary for arms control and international security, said. She said the U.S. political declaration, which contains non-legally binding guidelines outlining best practices for responsible military use of AI, "can be a focal point for international cooperation." Jenkins launched the declaration at the end of a two-day conference in The Hague that took on additional urgency as advances in drone technology amid the Russia's war in Ukraine have accelerated a trend that could soon bring the world's first fully autonomous fighting robots to the battlefield. The U.S. declaration has 12 points, including that military uses of AI are consistent with international law, and that states "maintain human control and involvement for all actions critical to informing and executing sovereign decisions concerning nuclear weapons employment."
- North America > United States (0.57)
- Europe > Ukraine (0.35)
- Europe > Netherlands > South Holland > The Hague (0.32)
- (3 more...)
AI diplomacy: five recommendations to developing countries
AI has extraordinary potential and developing countries must move forward quickly in this field to leverage their technological prowess, productivity, and competitiveness. Certainly, investing in R&D, developing capacities, and retaining AI talent is much easier said than done. Besides adopting a national AI strategy, if there is none, developing countries could put into practice a roadmap with clearly defined priorities and projects that bolster the economy. They can also build partnerships and reach out to other countries and organizations that are willing to cooperate in frontier technologies. A niche strategy might help to leapfrog in a few select sectors, as in the case of some small states that have become active players in the digital sphere. Interestingly enough, Kenya became last August the first African country to teach coding as a subject in schools. As stated in the UNCTAD 2021 Digital Economy report, developing countries risk becoming mere providers of data, while having to pay for digital intelligence produced with their data. Current international regulatory frameworks tend to be either too narrow in scope or too limited geographically, failing to enable cross-border data flows with an equitable sharing of economic gains. In a nutshell, developing countries need to find the optimal balance between promoting domestic economic development, protecting public policy interests, and integrating into the global digital ecosystem.
- Africa > Kenya (0.25)
- North America > United States > California > San Francisco County > San Francisco (0.05)